Memorization and generalization in neural code intelligence models
نویسندگان
چکیده
Deep Neural Networks (DNNs) are increasingly being used in software engineering and code intelligence tasks. These powerful tools that capable of learning highly generalizable patterns from large datasets through millions parameters. At the same time, their capacity can render them prone to memorizing data points. Recent work suggests memorization risk manifests especially strongly when training dataset is noisy, involving many ambiguous or questionable samples, only recourse. The goal this paper evaluate compare extent generalization neural models. It aims provide insights on how may impact behavior models systems. To observe models, we add random noise original use various metrics quantify aspects testing. We several state-of-the-art benchmarks based Java, Python, Ruby codebases. Our results highlight important risks: trainable parameters allow networks memorize anything, including noisy data, a false sense generalization. observed all manifest some forms memorization. This be potentially troublesome most tasks where they rely rather noise-prone repetitive sources, such as GitHub. best our knowledge, first study effects domain raises awareness provides new into issues systems usually overlooked by researchers.
منابع مشابه
Latent Geometry and Memorization in Generative Models
It can be difficult to tell whether a trained generative model has learned to generate novel examples or has simply memorized a specific set of outputs. In published work, it is common to attempt to address this visually, for example by displaying a generated example and its nearest neighbor(s) in the training set (in, for example, the L metric). As any generative model induces a probability de...
متن کاملGeneralization Performance of Regularized Neural Network Models
Architecture optimization is a fundamental problem of neural network modeling. The optimal architecture is defined as the one which minimizes the generalization error. This paper addresses estimation of the generalization performance of regularized, complete neural network models. Regularization normally improves the generalization performance by restricting the model complexity. A formula for ...
متن کاملUsing emotional intelligence to predict job stress: Artificial neural network and regression models
Introduction: These days, there is a consensus that emotional intelligence plays an important role in the success of individuals in different areas of life. Persons with higher emotional intelligence had lower stress in dealing with demands and pressures in the workplace. The purpose of this study was to use artificial neural network to predict job stress and to compare the performance of this ...
متن کاملMemorization and Association on a Realistic Neural Model
A central open question of computational neuroscience is to identify the data structures and algorithms that are used in mammalian cortex to support successive acts of the basic cognitive tasks of memorization and association. This letter addresses the simultaneous challenges of realizing these two distinct tasks with the same data structure, and doing so while respecting the following four bas...
متن کاملThe Secret Sharer: Measuring Unintended Neural Network Memorization & Extracting Secrets
Machine learning models based on neural networks and deep learning are being rapidly adopted for many purposes. What those models learn, and what they may share, is a significant concern when the training data may contain secrets and the models are public—e.g., when a model helps users compose text messages using models trained on all users’ messages. This paper presents exposure, a simple-to-c...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Information & Software Technology
سال: 2023
ISSN: ['0950-5849', '1873-6025']
DOI: https://doi.org/10.1016/j.infsof.2022.107066